Supervised Multivariate Learning with Simultaneous Feature Auto-Grouping and Dimension Reduction

نویسندگان

چکیده

Abstract Modern high-dimensional methods often adopt the ‘bet on sparsity’ principle, while in supervised multivariate learning statisticians may face ‘dense’ problems with a large number of nonzero coefficients. This paper proposes novel clustered reduced-rank (CRL) framework that imposes two joint matrix regularizations to automatically group features constructing predictive factors. CRL is more interpretable than low-rank modelling and relaxes stringent sparsity assumption variable selection. In this paper, new information-theoretical limits are presented reveal intrinsic cost seeking for clusters, as well blessing from dimensionality learning. Moreover, an efficient optimization algorithm developed, which performs subspace clustering guaranteed convergence. The obtained fixed-point estimators, although not necessarily globally optimal, enjoy desired statistical accuracy beyond standard likelihood setup under some regularity conditions. kind information criterion, its scale-free form, proposed cluster rank selection, has rigorous theoretical support without assuming infinite sample size. Extensive simulations real-data experiments demonstrate interpretability method.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Supervised dimension reduction mappings

Abstract. We propose a general principle to extend dimension reduction tools to explicit dimension reduction mappings and we show that this can serve as an interface to incorporate prior knowledge in the form of class labels. We explicitly demonstrate this technique by combining locally linear mappings which result from matrix learning vector quantization schemes with the t-distributed stochast...

متن کامل

Intrusion Detection using Supervised Learning with Feature Set Reduction

Intrusion detection systems intend to recognize attacks with a low false positive rate and high detection rate. Many feature selection methods introduced to eliminate redundant and irrelevant features, because raw features may abbreviate accuracy or robustness of classification. In this paper we are proposing the information gain technique for the selection of the features. A feature with the h...

متن کامل

Supervised dimension reduction with topic models

We consider supervised dimension reduction (SDR) for problems with discrete variables. Existing methods are computationally expensive, and often do not take the local structure of data into consideration when searching for a low-dimensional space. In this paper, we propose a novel framework for SDR which is (1) general and flexible so that it can be easily adapted to various unsupervised topic ...

متن کامل

Gradient-based kernel dimension reduction for supervised learning

This paper proposes a novel kernel approach to linear dimension reduction for supervised learning. The purpose of the dimension reduction is to find directions in the input space to explain the output as effectively as possible. The proposed method uses an estimator for the gradient of regression function, based on the covariance operators on reproducing kernel Hilbert spaces. In comparison wit...

متن کامل

Linear Contour Learning: A Method for Supervised Dimension Reduction

We propose a novel approach to sufficient di­ mension reduction in regression, based on es­ timating contour directions of negligible vari­ ation for the response surface. These di­ rections span the orthogonal complement of the minimal space relevant for the regression, and can be extracted according to a mea­ sure of the variation in the response, lead­ ing to General Contour Regression (GCR)...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Journal of The Royal Statistical Society Series B-statistical Methodology

سال: 2022

ISSN: ['1467-9868', '1369-7412']

DOI: https://doi.org/10.1111/rssb.12492